Convex polyhedron learning and its applications
نویسنده
چکیده
From a possible engineer’s point of view learning can be considered as discovering the relationship between the features of a phenomenon. Machine learning (data mining) is a variant of learning, in which the observations about the phenomenon are available as data, and the connection between the features is discovered by a program. In the case of classification the phenomenon is modeled by a random pair (X, Y ), where the d-dimensional continuous X is called input, and the discrete (often binary) Y is called label. In the case of collaborative filtering the phenomenon is modeled by a random triplet (U, I,R), where the discrete U is called user identifier, the discrete I is called item identifier, and the continuous R is called rating value. Unbalanced problems i.e. in which one class label occurs much more frequently than the other form an interesting subset among binary classification problems. In practice such problems often arise for example in the field of medical and technical diagnostics. A convex polyhedron classifier is a function g : R 7→ {+1,−1} with the property that the decision region {x ∈ R : g(x) = 1} is a convex polyhedron. At classifying an input x ∈ R, we have to substitute x into the linear functions defining the polyhedron. If any of the substitutions gives a negative number, then we can stop the computation immediately, since the class label will be necessarily −1 in this case. As a consequence, convex polyhedron classifiers fit well to unbalanced problems. The convex polyhedron based approach has its analogous variant for collaborative filtering too. In this case the utility of the approach is that it gives a unique solution of the problem that can be a useful component of a blended solution involving many different models. A related problem to classification is determining the convex separability of point sets. Let us assume that P and Q are finite sets in R. The task is to decide whether there exist a convex polyhedron S that contains all element of P, but no elements from Q. In a practical data mining project typically many experiments are run and many models are built. It is non-trivial to decide which of them should be used for prediction in the final system. Obviously, if two models achieve the same accuracy on the training set, then it is reasonable to choose the simpler one. The Vapnik–Chervonenkis dimension is a widely accepted model complexity measure in the case of binary classification. The first chapter of the thesis (Introduction) briefly introduces the field of machine learning and locates convex polyhedron learning in it. Then, without completeness it overviews a set known learning algorithms. The part dealing with collaborative filtering contains novel results too. The second chapter of the thesis (Algorithms) is about algorithms that use convex polyhedrons for solving various machine learning problems. The first part of the chapter deals with the problem of linear and convex separation. The second part of the chapter gives algorithms for training convex polyhedron classifiers. The third part of the chapter introduces a convex polyhedron based algorithm for collaborative filtering. The third chapter of the thesis (Model complexity) collects the known facts about the Vapnik– Chervonenkis dimension of convex polyhedron classifiers and proves new results. The fourth chapter (Applications) presents the experiments performed with the proposed algorithms.
منابع مشابه
Applications of strong convexity--strong smoothness duality to learning with matrices
It is known that a function is strongly convex with respect to some norm if and only if its conjugate function is strongly smooth with respect to the dual norm. This result has already been found to be a key component in deriving and analyzing several learning algorithms. Utilizing this duality, we isolate a single inequality which seamlessly implies both generalization bounds and online regret...
متن کاملStructured sparsity-inducing norms through submodular functions
Sparse methods for supervised learning aim at finding good linear predictors from as few variables as possible, i.e., with small cardinality of their supports. This combinatorial selection problem is often turned into a convex optimization problem by replacing the cardinality function by its convex envelope (tightest convex lower bound), in this case the l1-norm. In this paper, we investigate m...
متن کاملOn the duality of strong convexity and strong smoothness: Learning applications and matrix regularization
We show that a function is strongly convex with respect to some norm if and only if its conjugate function is strongly smooth with respect to the dual norm. This result has already been found to be a key component in deriving and analyzing several learning algorithms. Utilizing this duality, we isolate a single inequality which seamlessly implies both generalization bounds and online regret bou...
متن کاملLearning with Submodular Functions: A Convex Optimization Perspective
Submodular functions are relevant to machine learning for at least two reasons: (1) some problems may be expressed directly as the optimization of submodular functions and (2) the Lovász extension of submodular functions provides a useful set of regularization functions for supervised and unsupervised learning. In this monograph, we present the theory of submodular functions from a convex analy...
متن کاملDisjoint convex shell and its applications in mesh unfolding
In this work, we study a geometric structure called disjoint convex shell or simply DC-shell. A DC-shell of a polyhedron is a set of pairwise interior disjoint convex objects that collectively approximate the given polyhedron. Preventing convex objects from overlapping enables faster and robust collision response and more realistic fracturing simulation. Without the disjointness constraint, a p...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2009